A guided Minimax-DDQN (Minimax-Double Deep Q-Network) algorithm was designed to solve the problems of unpredictable enemy aircraft maneuver strategy and low winning rate, which are caused by the complex environment information and strong confrontation of Unmanned Aerial Vehicle (UAV) in air combat. Firstly, on the basis of Minimax decision-making method, a guided strategy exploration mechanism was proposed. Then, combined with the guided Minimax strategy, a type of DDQN (Double Deep Q-Network) algorithm was designed to improve the update efficiency of Q-network. Finally, an advanced three-stage network training method was proposed. And through confrontation training between different decision models, better optimized decision model was obtained. Experimental results show that compared with Minimax-DQN (Minimax-DQN), Minimax-DDQN and other algorithms, the proposed algorithm has the success rate of chasing straight target improved by 14% to 60% and the winning rate against DDQN algorithm over 60%. It can be seen that compared with algorithms such as DDQN and Minimax-DDQN, the proposed algorithm has stronger decision-making capability and better adaptability in high confrontation combat environment.